5,956 research outputs found
DPC-Net: Deep Pose Correction for Visual Localization
We present a novel method to fuse the power of deep networks with the
computational efficiency of geometric and probabilistic localization
algorithms. In contrast to other methods that completely replace a classical
visual estimator with a deep network, we propose an approach that uses a
convolutional neural network to learn difficult-to-model corrections to the
estimator from ground-truth training data. To this end, we derive a novel loss
function for learning SE(3) corrections based on a matrix Lie groups approach,
with a natural formulation for balancing translation and rotation errors. We
use this loss to train a Deep Pose Correction network (DPC-Net) that predicts
corrections for a particular estimator, sensor and environment. Using the KITTI
odometry dataset, we demonstrate significant improvements to the accuracy of a
computationally-efficient sparse stereo visual odometry pipeline, that render
it as accurate as a modern computationally-intensive dense estimator. Further,
we show how DPC-Net can be used to mitigate the effect of poorly calibrated
lens distortion parameters.Comment: In IEEE Robotics and Automation Letters (RA-L) and presented at the
IEEE International Conference on Robotics and Automation (ICRA'18), Brisbane,
Australia, May 21-25, 201
How to Train a CAT: Learning Canonical Appearance Transformations for Direct Visual Localization Under Illumination Change
Direct visual localization has recently enjoyed a resurgence in popularity
with the increasing availability of cheap mobile computing power. The
competitive accuracy and robustness of these algorithms compared to
state-of-the-art feature-based methods, as well as their natural ability to
yield dense maps, makes them an appealing choice for a variety of mobile
robotics applications. However, direct methods remain brittle in the face of
appearance change due to their underlying assumption of photometric
consistency, which is commonly violated in practice. In this paper, we propose
to mitigate this problem by training deep convolutional encoder-decoder models
to transform images of a scene such that they correspond to a previously-seen
canonical appearance. We validate our method in multiple environments and
illumination conditions using high-fidelity synthetic RGB-D datasets, and
integrate the trained models into a direct visual localization pipeline,
yielding improvements in visual odometry (VO) accuracy through time-varying
illumination conditions, as well as improved metric relocalization performance
under illumination change, where conventional methods normally fail. We further
provide a preliminary investigation of transfer learning from synthetic to real
environments in a localization context. An open-source implementation of our
method using PyTorch is available at https://github.com/utiasSTARS/cat-net.Comment: In IEEE Robotics and Automation Letters (RA-L) and presented at the
IEEE International Conference on Robotics and Automation (ICRA'18), Brisbane,
Australia, May 21-25, 201
Improving Foot-Mounted Inertial Navigation Through Real-Time Motion Classification
We present a method to improve the accuracy of a foot-mounted,
zero-velocity-aided inertial navigation system (INS) by varying estimator
parameters based on a real-time classification of motion type. We train a
support vector machine (SVM) classifier using inertial data recorded by a
single foot-mounted sensor to differentiate between six motion types (walking,
jogging, running, sprinting, crouch-walking, and ladder-climbing) and report
mean test classification accuracy of over 90% on a dataset with five different
subjects. From these motion types, we select two of the most common (walking
and running), and describe a method to compute optimal zero-velocity detection
parameters tailored to both a specific user and motion type by maximizing the
detector F-score. By combining the motion classifier with a set of optimal
detection parameters, we show how we can reduce INS position error during mixed
walking and running motion. We evaluate our adaptive system on a total of 5.9
km of indoor pedestrian navigation performed by five different subjects moving
along a 130 m path with surveyed ground truth markers.Comment: In Proceedings of the International Conference on Indoor Positioning
and Indoor Navigation (IPIN'17), Sapporo, Japan, Sep. 18-21, 201
Reference frames during the acquisition and development of spatial memories
Four experiments investigated the role of reference frames during the acquisition and development of spatial knowledge, when learning occurs incrementally across views. In two experiments, participants learned overlapping spatial layouts. Layout 1 was first studied in isolation, and Layout 2 was later studied in the presence of Layout 1. The Layout 1 learning view was manipulated, whereas the Layout 2 view was held constant. Manipulation of the Layout 1 view influenced the reference frame used to organize Layout 2, indicating that reference frames established during early environmental exposure provided a framework for organizing locations learned later. Further experiments demonstrated that reference frames established after learning served to reorganize an existing spatial memory. These results indicate that existing reference frames can structure the acquisition of new spatial memories and that new reference frames can reorganize existing spatial memories
Walking through a virtual environment improves perceived size within and beyond the walked space
Distances tend to be underperceived in virtual environments (VEs) by up to 50%, whereas distances tend to be perceived accurately in the real world. Previous work has shown that allowing participants to interact with the VE while receiving continual visual feedback can reduce this underperception. Judgments of virtual object size have been used to measure whether this improvement is due to the rescaling of perceived space, but there is disagreement within the literature as to whether judgments of object size benefit from interaction with feedback. This study contributes to that discussion by employing a more natural measure of object size. We also examined whether any improvement in virtual distance perception was limited to the space used for interaction (1â5 m) or extended beyond (7â11 m). The results indicated that object size judgments do benefit from interaction with the VE, and that this benefit extends to distances beyond the explored space
Online Thermal Analysis of Batch Roasted Coffee Beans
We constructed and instrumented a fluidised-bed coffee roaster. This work has been carried out as part of a search for the âideal pointâ, which is the point in time when an expert roaster would terminate the roast in order to yield beans that produce the optimal brew. We roasted Costa Rican Arabica beans whilst controlling the roasting temperature to follow a linear ramp. We measured and recorded the input, output, and coffee bean surface temperatures. We introduce the idea of âbean loadâ, an uncalibrated measure of the heat load presented by the material being roasted. The bean load under constantly-ramping bean surface temperature shows the roast is increasingly endothermic. Toward the end of the roast the endothermic phenomena decrease, or are assisted by exothermic activity. The bean load also has a repeatable dip around first crack. Due to limitations with the roaster we were not able to make reliable measurements at and beyond second crack. We observed no waypoints or events that might be used to pinpoint the âideal pointâ to end the roast
Learning Matchable Image Transformations for Long-term Metric Visual Localization
Long-term metric self-localization is an essential capability of autonomous
mobile robots, but remains challenging for vision-based systems due to
appearance changes caused by lighting, weather, or seasonal variations. While
experience-based mapping has proven to be an effective technique for bridging
the `appearance gap,' the number of experiences required for reliable metric
localization over days or months can be very large, and methods for reducing
the necessary number of experiences are needed for this approach to scale.
Taking inspiration from color constancy theory, we learn a nonlinear
RGB-to-grayscale mapping that explicitly maximizes the number of inlier feature
matches for images captured under different lighting and weather conditions,
and use it as a pre-processing step in a conventional single-experience
localization pipeline to improve its robustness to appearance change. We train
this mapping by approximating the target non-differentiable localization
pipeline with a deep neural network, and find that incorporating a learned
low-dimensional context feature can further improve cross-appearance feature
matching. Using synthetic and real-world datasets, we demonstrate substantial
improvements in localization performance across day-night cycles, enabling
continuous metric localization over a 30-hour period using a single mapping
experience, and allowing experience-based localization to scale to long
deployments with dramatically reduced data requirements.Comment: In IEEE Robotics and Automation Letters (RA-L) and presented at the
IEEE International Conference on Robotics and Automation (ICRA'20), Paris,
France, May 31-June 4, 202
The Phoenix Drone: An Open-Source Dual-Rotor Tail-Sitter Platform for Research and Education
In this paper, we introduce the Phoenix drone: the first completely
open-source tail-sitter micro aerial vehicle (MAV) platform. The vehicle has a
highly versatile, dual-rotor design and is engineered to be low-cost and easily
extensible/modifiable. Our open-source release includes all of the design
documents, software resources, and simulation tools needed to build and fly a
high-performance tail-sitter for research and educational purposes. The drone
has been developed for precision flight with a high degree of control
authority. Our design methodology included extensive testing and
characterization of the aerodynamic properties of the vehicle. The platform
incorporates many off-the-shelf components and 3D-printed parts, in order to
keep the cost down. Nonetheless, the paper includes results from flight trials
which demonstrate that the vehicle is capable of very stable hovering and
accurate trajectory tracking. Our hope is that the open-source Phoenix
reference design will be useful to both researchers and educators. In
particular, the details in this paper and the available open-source materials
should enable learners to gain an understanding of aerodynamics, flight
control, state estimation, software design, and simulation, while experimenting
with a unique aerial robot.Comment: In Proceedings of the IEEE International Conference on Robotics and
Automation (ICRA'19), Montreal, Canada, May 20-24, 201
- âŠ